Differential entropy (also referred to as continuous entropy) is a concept in information theory that extends the idea of (Shannon) entropy, a measure of average surprisal of a random variable, to continuous probability distributions.
Contents |
Let X be a random variable with a probability density function f whose support is a set . The differential entropy or is defined as
As with its discrete analog, the units of differential entropy depend on the base of the logarithm, which is usually 2 (i.e., the units are bits). See logarithmic units for logarithms taken in different bases. Related concepts such as joint, conditional differential entropy, and relative entropy are defined in a similar fashion.
One must take care in trying to apply properties of discrete entropy to differential entropy, since probability density functions can be greater than 1. For example, Uniform(0,1/2) has negative differential entropy .
Thus, differential entropy does not share all properties of discrete entropy.
Note that the continuous mutual information has the distinction of retaining its fundamental significance as a measure of discrete information since it is actually the limit of the discrete mutual information of partitions of X and Y as these partitions become finer and finer. Thus it is invariant under non-linear homeomorphisms (continuous and uniquely invertible maps) ,[1] including linear [2] transformations of X and Y, and still represents the amount of discrete information that can be transmitted over a channel that admits a continuous space of values.
However, differential entropy does not have other desirable properties:
A modification of differential entropy that addresses this is the relative information entropy, also known as the Kullback–Leibler divergence, which includes an invariant measure factor (see limiting density of discrete points).
With a normal distribution, differential entropy is maximized for a given variance. The following is a proof that a Gaussian variable has the largest entropy amongst all random variables of equal variance.
Let be a Gaussian PDF with mean and variance and an arbitrary PDF with the same variance. Since differential entropy is translation invariant we can assume that has the same mean of as .
Consider the Kullback-Leibler divergence between the two distributions
Now note that
because the result does not depend on other than through the variance. Combining the two results yields
with equality when following from the properties of Kullback-Leibler divergence.
This result may also be demonstrated using the variational calculus. A Lagrangian function with two Lagrangian multipliers may be defined as:
where g(x) is some function with mean μ. When the entropy of g(x) is at a maximum and the constraint equations, which consist of the normalization condition and the requirement of fixed variance , are both satisfied, then a small variation about g(x) will produce a variation about L which is equal to zero:
Since this must hold for any small , the term in brackets must be zero, and solving for g(x) yields:
Using the constraint equations to solve for and yields the normal distribution:
Let X be an exponentially distributed random variable with parameter , that is, with probability density function
Its differential entropy is then
Here, was used rather than to make it explicit that the logarithm was taken to base e, to simplify the calculation.
In the table below, (the gamma function), , , and is Euler's constant. Each distribution maximizes the entropy for a particular set of functional constraints listed in the fourth column, and the constraint that x be included in the support of the probability density, which is listed in the fifth column.[3]
Distribution Name | Probability density function (pdf) | Entropy in nats | Maximum Entropy Constraint | Support |
---|---|---|---|---|
Uniform | None | |||
Normal | ||||
Exponential | ||||
Rayleigh | ||||
Beta | for | |||
Cauchy | ||||
Chi | ||||
Chi-squared | ||||
Erlang | ||||
F | ||||
Gamma | ||||
Laplace | ||||
Logistic | ||||
Lognormal | ||||
Maxwell-Boltzmann | ||||
Generalized normal | ||||
Pareto | ||||
Student's t | ||||
Triangular | ||||
Weibull | ||||
Multivariate normal |
(Many of the differential entropies are from.[4]
As described above, differential entropy does not share all properties of discrete entropy. A modification of differential entropy adds an invariant measure factor to correct this, (see limiting density of discrete points). If m(x) is further constrained to be a probability density, the resulting notion is called relative entropy in information theory:
The definition of differential entropy above can be obtained by partitioning the range of X into bins of length with associated sample points ih within the bins, for X Riemann integrable. This gives a quantized version of X, defined by if . Then the entropy of is
The first term on the right approximates the differential entropy, while the second term is approximately . Note that this procedure suggests that the entropy in the discrete sense of a continuous random variable should be .